Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 11.164
Filtrar
1.
J Nurs Educ ; 63(5): 298-303, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38729141

RESUMO

BACKGROUND: This randomized controlled, pretest-post-test intervention study examined the effect of distance reiki on state test anxiety and test performance. METHOD: First-year nursing students (n = 71) were randomized into two groups. One week before the examination, intervention group participants performed reiki remotely for 20 minutes for 4 consecutive days, and control group participants received no intervention. RESULTS: The intervention group had lower posttest cognitive and psychosocial subscale scores than pretest scores (p > .05). The control group had a significantly higher mean posttest physiological subscale score than pretest score (p < .05). Final grade point averages were not significantly different between the intervention and control groups (p > .05). One quarter of the intervention group participants noted reiki reduced their stress and helped them perform better on the examination. CONCLUSION: Reiki is a safe and easy-to-practice method to help students cope with test anxiety. [J Nurs Educ. 2024;63(5):298-303.].


Assuntos
Estudantes de Enfermagem , Humanos , Estudantes de Enfermagem/psicologia , Estudantes de Enfermagem/estatística & dados numéricos , Feminino , Masculino , Adulto Jovem , Avaliação Educacional/métodos , Bacharelado em Enfermagem/métodos , Toque Terapêutico , Ansiedade aos Exames , Adulto , Ansiedade/prevenção & controle
2.
Saudi Med J ; 45(5): 531-536, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38734438

RESUMO

OBJECTIVES: To evaluate the role of artificial intelligence (Google Bard) in figures, scans, and image identifications and interpretations in medical education and healthcare sciences through an Objective Structured Practical Examination (OSPE) type of performance. METHODS: The OSPE type of question bank was created with a pool of medical sciences figures, scans, and images. For assessment, 60 figures, scans and images were selected and entered into the given area of the Google Bard to evaluate the knowledge level. RESULTS: The marks obtained by Google Bard in brain structures, morphological and radiological images 7/10 (70%); bone structures, radiological images 9/10 (90%); liver structure and morphological, pathological images 4/10 (40%); kidneys structure and morphological images 2/7 (28.57%); neuro-radiological images 4/7 (57.14%); and endocrine glands including the thyroid, pancreas, breast morphological and radiological images 8/16 (50%). The overall total marks obtained by Google Bard in various OSPE figures, scans, and image identification questions were 34/60 (56.7%). CONCLUSION: Google Bard scored satisfactorily in morphological, histopathological, and radiological image identifications and their interpretations. Google Bard may assist medical students, faculty in medical education and physicians in healthcare settings.


Assuntos
Inteligência Artificial , Humanos , Educação Médica/métodos , Avaliação Educacional/métodos , Radiografia/métodos
3.
BMC Med Educ ; 24(1): 527, 2024 May 11.
Artigo em Inglês | MEDLINE | ID: mdl-38734603

RESUMO

BACKGROUND: High stakes examinations used to credential trainees for independent specialist practice should be evaluated periodically to ensure defensible decisions are made. This study aims to quantify the College of Intensive Care Medicine of Australia and New Zealand (CICM) Hot Case reliability coefficient and evaluate contributions to variance from candidates, cases and examiners. METHODS: This retrospective, de-identified analysis of CICM examination data used descriptive statistics and generalisability theory to evaluate the reliability of the Hot Case examination component. Decision studies were used to project generalisability coefficients for alternate examination designs. RESULTS: Examination results from 2019 to 2022 included 592 Hot Cases, totalling 1184 individual examiner scores. The mean examiner Hot Case score was 5.17 (standard deviation 1.65). The correlation between candidates' two Hot Case scores was low (0.30). The overall reliability coefficient for the Hot Case component consisting of two cases observed by two separate pairs of examiners was 0.42. Sources of variance included candidate proficiency (25%), case difficulty and case specificity (63.4%), examiner stringency (3.5%) and other error (8.2%). To achieve a reliability coefficient of > 0.8 a candidate would need to perform 11 Hot Cases observed by two examiners. CONCLUSION: The reliability coefficient for the Hot Case component of the CICM second part examination is below the generally accepted value for a high stakes examination. Modifications to case selection and introduction of a clear scoring rubric to mitigate the effects of variation in case difficulty may be helpful. Increasing the number of cases and overall assessment time appears to be the best way to increase the overall reliability. Further research is required to assess the combined reliability of the Hot Case and viva components.


Assuntos
Competência Clínica , Cuidados Críticos , Avaliação Educacional , Humanos , Nova Zelândia , Austrália , Reprodutibilidade dos Testes , Estudos Retrospectivos , Cuidados Críticos/normas , Avaliação Educacional/métodos , Educação de Pós-Graduação em Medicina/normas
4.
S Afr Fam Pract (2004) ; 66(1): e1-e15, 2024 Apr 26.
Artigo em Inglês | MEDLINE | ID: mdl-38708750

RESUMO

BACKGROUND:  Learning portfolios (LPs) provide evidence of workplace-based assessments (WPBAs) in clinical settings. The educational impact of LPs has been explored in high-income countries, but the use of portfolios and the types of assessments used for and of learning have not been adequately researched in sub-Saharan Africa. This study investigated the evidence of learning in registrars' LPs and the influence of the training district and year of training on assessments. METHODS:  A cross-sectional study evaluated 18 Family Medicine registrars' portfolios from study years 1-3 across five decentralised training sites affiliated with the University of the Witwatersrand. Descriptive statistics were calculated for the portfolio and quarterly assessment (QA) scores and self-reported clinical skills competence levels. The competence levels obtained from the portfolios and university records served as proxy measures for registrars' knowledge and skills. RESULTS:  The total LP median scores ranged from 59.9 to 81.0, and QAs median scores from 61.4 to 67.3 across training years. The total LP median scores ranged from 62.1 to 83.5 and 62.0 to 67.5, respectively in QAs across training districts. Registrars' competence levels across skill sets did not meet the required standards. Higher skills competence levels were reported in the women's health, child health, emergency care, clinical administration and teaching and learning domains. CONCLUSION:  The training district and training year influence workplace-based assessment (WPBA) effectiveness. Ongoing faculty development and registrar support are essential for WPBA.Contribution: This study contributes to the ongoing discussion of how to utilise WPBA in resource-constrained sub-Saharan settings.


Assuntos
Competência Clínica , Avaliação Educacional , Medicina de Família e Comunidade , Local de Trabalho , Humanos , Estudos Transversais , Medicina de Família e Comunidade/educação , Avaliação Educacional/métodos , Feminino , Masculino , África do Sul , Aprendizagem , Adulto
5.
BMC Med Educ ; 24(1): 502, 2024 May 09.
Artigo em Inglês | MEDLINE | ID: mdl-38724925

RESUMO

INTRODUCTION: The Clinical Skill Training Center (CSTC) is the first environment where third year medical students learn clinical skills after passing basic science. Consumer- based evaluation is one of the ways to improve this center with the consumer. This study was conducted with the aim of preparing a consumer-oriented evaluation tool for CSTC among medical students. METHOD: The study was mixed method. The first phase was qualitative and for providing an evaluation tool. The second phase was for evaluating the tool. At the first phase, after literature review in the Divergent phase, a complete list of problems in the field of CSTC in medicine schools was prepared. In the convergent step, the prepared list was compared with the standards of clinical education and values of scriven. In the second phase it was evaluated by the scientific and authority committee. Validity has been measured by determining CVR and CVI: Index. The face and content validity of the tool was obtained through the approval of a group of specialists. RESULTS: The findings of the research were in the form of 4 questionnaires: clinical instructors, pre-clinical medical students, and interns. All items were designed as a 5-point Likert. The main areas of evaluation included the objectives and content of training courses, implementation of operations, facilities and equipment, and the environment and indoor space. In order to examine the long-term effects, a special evaluation form was designed for intern. CONCLUSION: The tool for consumer evaluation was designed with good reliability and trustworthiness and suitable for use in the CSTC, and its use can improve the effectiveness of clinical education activities.


Assuntos
Competência Clínica , Avaliação de Programas e Projetos de Saúde , Estudantes de Medicina , Humanos , Competência Clínica/normas , Educação de Graduação em Medicina/normas , Inquéritos e Questionários , Avaliação Educacional/métodos
6.
BMJ Open Qual ; 13(Suppl 2)2024 May 07.
Artigo em Inglês | MEDLINE | ID: mdl-38719519

RESUMO

INTRODUCTION: Safe practice in medicine and dentistry has been a global priority area in which large knowledge gaps are present.Patient safety strategies aim at preventing unintended damage to patients that can be caused by healthcare practitioners. One of the components of patient safety is safe clinical practice. Patient safety efforts will help in ensuring safe dental practice for early detection and limiting non-preventable errors.A valid and reliable instrument is required to assess the knowledge of dental students regarding patient safety. OBJECTIVE: To determine the psychometric properties of a written test to assess safe dental practice in undergraduate dental students. MATERIAL AND METHODS: A test comprising 42 multiple-choice questions of one-best type was administered to final year students (52) of a private dental college. Items were developed according to National Board of Medical Examiners item writing guidelines. The content of the test was determined in consultation with dental experts (either professor or associate professor). These experts had to assess each item on the test for language clarity as A: clear, B: ambiguous and relevance as 1: essential, 2: useful, not necessary, 3: not essential. Ethical approval was taken from the concerned dental college. Statistical analysis was done in SPSS V.25 in which descriptive analysis, item analysis and Cronbach's alpha were measured. RESULT: The test scores had a reliability (calculated by Cronbach's alpha) of 0.722 before and 0.855 after removing 15 items. CONCLUSION: A reliable and valid test was developed which will help to assess the knowledge of dental students regarding safe dental practice. This can guide medical educationist to develop or improve patient safety curriculum to ensure safe dental practice.


Assuntos
Avaliação Educacional , Segurança do Paciente , Psicometria , Humanos , Psicometria/instrumentação , Psicometria/métodos , Segurança do Paciente/normas , Segurança do Paciente/estatística & dados numéricos , Inquéritos e Questionários , Avaliação Educacional/métodos , Avaliação Educacional/estatística & dados numéricos , Avaliação Educacional/normas , Reprodutibilidade dos Testes , Estudantes de Odontologia/estatística & dados numéricos , Estudantes de Odontologia/psicologia , Educação em Odontologia/métodos , Educação em Odontologia/normas , Masculino , Feminino , Competência Clínica/estatística & dados numéricos , Competência Clínica/normas
8.
MedEdPORTAL ; 20: 11401, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38716162

RESUMO

Introduction: Vascular anomalies are a spectrum of disorders, including vascular tumors and malformations, that often require multispecialty care. The rarity and variety of these lesions make diagnosis, treatment, and management challenging. Despite the recognition of the medical complexity and morbidity associated with vascular anomalies, there is a general lack of education on the subject for pediatric primary care and subspecialty providers. A needs assessment and the lack of an available standardized teaching tool presented an opportunity to create an educational workshop for pediatric trainees using the POGIL (process-oriented guided inquiry learning) framework. Methods: We developed a 2-hour workshop consisting of an introductory didactic followed by small- and large-group collaboration and case-based discussion. The resource included customizable content for learning assessment and evaluation. Residents completed pre- and posttest assessments of content and provided written evaluations of the teaching session. Results: Thirty-four learners in pediatrics participated in the workshop. Session evaluations were positive, with Likert responses of 4.6-4.8 out of 5 on all items. Pre- and posttest comparisons of four content questions showed no overall statistically significant changes in correct response rates. Learners indicated plans to use the clinical content in their practice and particularly appreciated the interactive teaching forum and the comprehensive overview of vascular anomalies. Discussion: Vascular anomalies are complex, potentially morbid, and often lifelong conditions; multispecialty collaboration is key to providing comprehensive care for affected patients. This customizable resource offers a framework for trainees in pediatrics to appropriately recognize, evaluate, and refer patients with vascular anomalies.


Assuntos
Hemangioma , Internato e Residência , Pediatria , Malformações Vasculares , Humanos , Pediatria/educação , Pediatria/métodos , Internato e Residência/métodos , Malformações Vasculares/diagnóstico , Hemangioma/diagnóstico , Ensino , Aprendizagem Baseada em Problemas/métodos , Avaliação Educacional/métodos , Educação de Pós-Graduação em Medicina/métodos , Currículo
10.
J Coll Physicians Surg Pak ; 34(5): 614-616, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38720226

RESUMO

College of Physicians and Surgeons, Pakistan (CPSP) is a premier postgraduate medical institution of the country. It introduced Objective Structured Clinical Examination (OSCE) in the 1990s, and later came up with its modified form known as Task Oriented Assessment of Clinical Skills (TOACS). This modified assessment has been incorporated in clinical examinations of its majority fellowship programmes. Despite the use of TOACS for so many years at CPSP, it is surprising to note that this form of assessment does not appear in the literature. The objective of this viewpoint is to describe the rationale for the development of TOACS and to compare its structure and functions with OSCE. Key Words: Medical education, Assessment, Objective Structured Clinical Examination, Interactive, Task Oriented Assessment of Clinical Skills.


Assuntos
Competência Clínica , Avaliação Educacional , Humanos , Avaliação Educacional/métodos , Paquistão , Educação de Pós-Graduação em Medicina/métodos
11.
J Coll Physicians Surg Pak ; 34(5): 595-599, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38720222

RESUMO

OBJECTIVE: To analyse and compare the assessment and grading of human-written and machine-written formative essays. STUDY DESIGN: Quasi-experimental, qualitative cross-sectional study. Place and Duration of the Study: Department of Science of Dental Materials, Hamdard College of Medicine & Dentistry, Hamdard University, Karachi, from February to April 2023. METHODOLOGY: Ten short formative essays of final-year dental students were manually assessed and graded. These essays were then graded using ChatGPT version 3.5. The chatbot responses and prompts were recorded and matched with manually graded essays. Qualitative analysis of the chatbot responses was then performed. RESULTS: Four different prompts were given to the artificial intelligence (AI) driven platform of ChatGPT to grade the summative essays. These were the chatbot's initial responses without grading, the chatbot's response to grading against criteria, the chatbot's response to criteria-wise grading, and the chatbot's response to questions for the difference in grading. Based on the results, four innovative ways of using AI and machine learning (ML) have been proposed for medical educators: Automated grading, content analysis, plagiarism detection, and formative assessment. ChatGPT provided a comprehensive report with feedback on writing skills, as opposed to manual grading of essays. CONCLUSION: The chatbot's responses were fascinating and thought-provoking. AI and ML technologies can potentially supplement human grading in the assessment of essays. Medical educators need to embrace AI and ML technology to enhance the standards and quality of medical education, particularly when assessing long and short essay-type questions. Further empirical research and evaluation are needed to confirm their effectiveness. KEY WORDS: Machine learning, Artificial intelligence, Essays, ChatGPT, Formative assessment.


Assuntos
Inteligência Artificial , Avaliação Educacional , Aprendizado de Máquina , Humanos , Estudos Transversais , Avaliação Educacional/métodos , Paquistão , Educação Médica/métodos , Estudantes de Odontologia/psicologia , Redação , Pesquisa Qualitativa , Educação em Odontologia/métodos
12.
BMC Med Educ ; 24(1): 504, 2024 May 07.
Artigo em Inglês | MEDLINE | ID: mdl-38714975

RESUMO

BACKGROUND: Evaluation of students' learning strategies can enhance academic support. Few studies have investigated differences in learning strategies between male and female students as well as their impact on United States Medical Licensing Examination® (USMLE) Step 1 and preclinical performance. METHODS: The Learning and Study Strategies Inventory (LASSI) was administered to the classes of 2019-2024 (female (n = 350) and male (n = 262)). Students' performance on preclinical first-year (M1) courses, preclinical second-year (M2) courses, and USMLE Step 1 was recorded. An independent t-test evaluated differences between females and males on each LASSI scale. A Pearson product moment correlation determined which LASSI scales correlated with preclinical performance and USMLE Step 1 examinations. RESULTS: Of the 10 LASSI scales, Anxiety, Attention, Information Processing, Selecting Main Idea, Test Strategies and Using Academic Resources showed significant differences between genders. Females reported higher levels of Anxiety (p < 0.001), which significantly influenced their performance. While males and females scored similarly in Concentration, Motivation, and Time Management, these scales were significant predictors of performance variation in females. Test Strategies was the largest contributor to performance variation for all students, regardless of gender. CONCLUSION: Gender differences in learning influence performance on STEP1. Consideration of this study's results will allow for targeted interventions for academic success.


Assuntos
Educação de Graduação em Medicina , Avaliação Educacional , Licenciamento em Medicina , Estudantes de Medicina , Humanos , Feminino , Masculino , Avaliação Educacional/métodos , Educação de Graduação em Medicina/normas , Fatores Sexuais , Licenciamento em Medicina/normas , Aprendizagem , Estados Unidos , Desempenho Acadêmico , Adulto Jovem
13.
JAMA Netw Open ; 7(5): e2410127, 2024 May 01.
Artigo em Inglês | MEDLINE | ID: mdl-38713464

RESUMO

Importance: Board certification can have broad implications for candidates' career trajectories, and prior research has found sociodemographic disparities in pass rates. Barriers in the format and administration of the oral board examinations may disproportionately affect certain candidates. Objective: To characterize oral certifying examination policies and practices of the 16 Accreditation Council for Graduate Medical Education (ACGME)-accredited specialties that require oral examinations. Design, Setting, and Participants: This cross-sectional study was conducted from March 1 to April 15, 2023, using data on oral examination practices and policies (examination format, dates, and setting; lactation accommodations; and accommodations for military deployment, family emergency, or medical leave) as well as the gender composition of the specialties' boards of directors obtained from websites, telephone calls and email correspondence with certifying specialists. The percentages of female residents and residents of racial and ethnic backgrounds who are historically underrepresented in medicine (URM) in each specialty as of December 31, 2021, were obtained from the Graduate Medical Education 2021 to 2022 report. Main Outcome and Measures: For each specialty, accommodation scores were measured by a modified objective scoring system (score range: 1-13, with higher scores indicating more accommodations). Poisson regression was used to assess the association between accommodation score and the diversity of residents in that specialty, as measured by the percentages of female and URM residents. Linear regression was used to assess whether gender diversity of a specialty's board of directors was associated with accommodation scores. Results: Included in the analysis were 16 specialties with a total of 46 027 residents (26 533 males [57.6%]) and 233 members of boards of directors (152 males [65.2%]). The mean (SD) total accommodation score was 8.28 (3.79), and the median (IQR) score was 9.25 (5.00-12.00). No association was found between test accommodation score and the percentage of female or URM residents. However, for each 1-point increase in the test accommodation score, the relative risk that a resident was female was 1.05 (95% CI, 0.96-1.16), and the relative risk that an individual was a URM resident was 1.04 (95% CI, 1.00-1.07). An association was found between the percentage of female board members and the accommodation score: for each 10% increase in the percentage of board members who were female, the accommodation score increased by 1.20 points (95% CI, 0.23-2.16 points; P = .03). Conclusions and Relevance: This cross-sectional study found considerable variability in oral board examination accommodations among ACGME-accredited specialties, highlighting opportunities for improvement and standardization. Promoting diversity in leadership bodies may lead to greater accommodations for examinees in extenuating circumstances.


Assuntos
Certificação , Humanos , Estudos Transversais , Feminino , Masculino , Certificação/estatística & dados numéricos , Estados Unidos , Conselhos de Especialidade Profissional/estatística & dados numéricos , Avaliação Educacional/estatística & dados numéricos , Avaliação Educacional/métodos , Educação de Pós-Graduação em Medicina/estatística & dados numéricos , Medicina/estatística & dados numéricos , Adulto
14.
JMIR Med Educ ; 10: e55595, 2024 Apr 26.
Artigo em Inglês | MEDLINE | ID: mdl-38693697

RESUMO

Background: Artificial intelligence (AI) chatbots, such as ChatGPT-4, have shown immense potential for application across various aspects of medicine, including medical education, clinical practice, and research. Objective: This study aimed to evaluate the performance of ChatGPT-4 in the 2023 Taiwan Audiologist Qualification Examination, thereby preliminarily exploring the potential utility of AI chatbots in the fields of audiology and hearing care services. Methods: ChatGPT-4 was tasked to provide answers and reasoning for the 2023 Taiwan Audiologist Qualification Examination. The examination encompassed six subjects: (1) basic auditory science, (2) behavioral audiology, (3) electrophysiological audiology, (4) principles and practice of hearing devices, (5) health and rehabilitation of the auditory and balance systems, and (6) auditory and speech communication disorders (including professional ethics). Each subject included 50 multiple-choice questions, with the exception of behavioral audiology, which had 49 questions, amounting to a total of 299 questions. Results: The correct answer rates across the 6 subjects were as follows: 88% for basic auditory science, 63% for behavioral audiology, 58% for electrophysiological audiology, 72% for principles and practice of hearing devices, 80% for health and rehabilitation of the auditory and balance systems, and 86% for auditory and speech communication disorders (including professional ethics). The overall accuracy rate for the 299 questions was 75%, which surpasses the examination's passing criteria of an average 60% accuracy rate across all subjects. A comprehensive review of ChatGPT-4's responses indicated that incorrect answers were predominantly due to information errors. Conclusions: ChatGPT-4 demonstrated a robust performance in the Taiwan Audiologist Qualification Examination, showcasing effective logical reasoning skills. Our results suggest that with enhanced information accuracy, ChatGPT-4's performance could be further improved. This study indicates significant potential for the application of AI chatbots in audiology and hearing care services.


Assuntos
Inteligência Artificial , Audiologistas , Audiologia , Humanos , Taiwan , Audiologia/métodos , Avaliação Educacional/métodos , Masculino , Competência Clínica/normas , Feminino
15.
BMC Med Educ ; 24(1): 487, 2024 May 02.
Artigo em Inglês | MEDLINE | ID: mdl-38698352

RESUMO

BACKGROUND: Workplace-based assessment (WBA) used in post-graduate medical education relies on physician supervisors' feedback. However, in a training environment where supervisors are unavailable to assess certain aspects of a resident's performance, nurses are well-positioned to do so. The Ottawa Resident Observation Form for Nurses (O-RON) was developed to capture nurses' assessment of trainee performance and results have demonstrated strong evidence for validity in Orthopedic Surgery. However, different clinical settings may impact a tool's performance. This project studied the use of the O-RON in three different specialties at the University of Ottawa. METHODS: O-RON forms were distributed on Internal Medicine, General Surgery, and Obstetrical wards at the University of Ottawa over nine months. Validity evidence related to quantitative data was collected. Exit interviews with nurse managers were performed and content was thematically analyzed. RESULTS: 179 O-RONs were completed on 30 residents. With four forms per resident, the ORON's reliability was 0.82. Global judgement response and frequency of concerns was correlated (r = 0.627, P < 0.001). CONCLUSIONS: Consistent with the original study, the findings demonstrated strong evidence for validity. However, the number of forms collected was less than expected. Exit interviews identified factors impacting form completion, which included clinical workloads and interprofessional dynamics.


Assuntos
Competência Clínica , Internato e Residência , Psicometria , Humanos , Reprodutibilidade dos Testes , Feminino , Masculino , Avaliação Educacional/métodos , Ontário , Medicina Interna/educação
16.
BMC Med Educ ; 24(1): 367, 2024 Apr 03.
Artigo em Inglês | MEDLINE | ID: mdl-38570816

RESUMO

INTRODUCTION: Psychometrics plays a vital role in evaluating educational research, including the analysis of multiple-choice exams. This study aims to improve the discriminatory ability of the "Médico Interno Residente" (MIR) medical exam in Spain, used to rank candidates for specialized healthcare training, through psychometric analysis. METHODS: We analyzed 2,890 MIR exam questions from 2009 to 2021 (totaling 147,214 exams), categorizing them based on methodology and response type. Evaluation employed classical test theory and item response theory (IRT). Classical test theory determined difficulty and discrimination indices, while IRT assessed the relationship between knowledge levels and question performance. RESULTS: Question distribution varied across categories and years. Frequently addressed knowledge areas included various medical specialties. Non-image-associated clinical cases were the easiest, while case-based clinical questions exhibited the highest discriminatory capacity, differing significantly from image-based case or negative questions. High-quality questions without images had longer stems but shorter answer choices. Adding images reduced discriminatory power and question difficulty, with image-based questions being easier. Clinical cases with images had shorter stems and longer answer choices. CONCLUSIONS: For improved exam performance, we recommend using a clinical case format followed by direct short-answer questions. Questions should be of low difficulty, providing clear and specific answers based on scientific evidence and avoiding ambiguity. Typical clinical cases with key characteristic features should be presented, excluding uncertain boundaries of medical knowledge. Questions should have lengthy stems and concise answer choices, minimizing speculation. If images are used, they should be typical, clear, consistent with the exam, and presented within clinical cases using clinical semiotics and propaedeutics.


Assuntos
Educação de Graduação em Medicina , Internato e Residência , Medicina , Humanos , Avaliação Educacional/métodos , Espanha , Educação de Graduação em Medicina/métodos
17.
Sci Rep ; 14(1): 9330, 2024 04 23.
Artigo em Inglês | MEDLINE | ID: mdl-38654011

RESUMO

While there is data assessing the test performance of artificial intelligence (AI) chatbots, including the Generative Pre-trained Transformer 4.0 (GPT 4) chatbot (ChatGPT 4.0), there is scarce data on its diagnostic accuracy of clinical cases. We assessed the large language model (LLM), ChatGPT 4.0, on its ability to answer questions from the United States Medical Licensing Exam (USMLE) Step 2, as well as its ability to generate a differential diagnosis based on corresponding clinical vignettes from published case reports. A total of 109 Step 2 Clinical Knowledge (CK) practice questions were inputted into both ChatGPT 3.5 and ChatGPT 4.0, asking ChatGPT to pick the correct answer. Compared to its previous version, ChatGPT 3.5, we found improved accuracy of ChatGPT 4.0 when answering these questions, from 47.7 to 87.2% (p = 0.035) respectively. Utilizing the topics tested on Step 2 CK questions, we additionally found 63 corresponding published case report vignettes and asked ChatGPT 4.0 to come up with its top three differential diagnosis. ChatGPT 4.0 accurately created a shortlist of differential diagnoses in 74.6% of the 63 case reports (74.6%). We analyzed ChatGPT 4.0's confidence in its diagnosis by asking it to rank its top three differentials from most to least likely. Out of the 47 correct diagnoses, 33 were the first (70.2%) on the differential diagnosis list, 11 were second (23.4%), and three were third (6.4%). Our study shows the continued iterative improvement in ChatGPT's ability to answer standardized USMLE questions accurately and provides insights into ChatGPT's clinical diagnostic accuracy.


Assuntos
Inteligência Artificial , Humanos , Estados Unidos , Diagnóstico Diferencial , Licenciamento em Medicina , Competência Clínica , Avaliação Educacional/métodos
18.
Surg Innov ; 31(3): 318-323, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38596895

RESUMO

OBJECTIVE: The aim was to determine whether preclinical medical students can develop their illness scripts to a level comparable to that of clerkship students through test-only learning using repeated formative online testlets. METHODS: In this experimental study, participants were 52 preclinical and 53 clinical medical students. The intervention group consisted of preclinical medical students, and the control group consisted of clinical medical students. The intervention group responded to online testlets containing feedback, an innovative formative assessment method called ContExtended Questions, on general surgery for 8 days by spending no more than 30 minutes each day. The control group completed the general surgery clerkship. The performances were assessed using 20 Key-Feature Question items. The intervention group was assessed twice: immediately after the intervention (the immediate test), and again 1 month later (the delayed test). The control group was assessed once, immediately after the clerkship. All performance tests were identical. RESULTS: The preclinical students had a significantly higher mean score on the immediate test (83.1 ± 9.6) compared to the clinical students (75.4 ± 8.9), P < .001. The effect size (Cohen's d) was .83. However, the mean score in the delayed test (76.9 ± 13.6) was not significantly different from clinical students' mean score (75.4 ± 8.9), P > .05. CONCLUSIONS: Test-only learning as a spaced repetition of online formative testlets is effective in preparing preclinical medical students to the clinical clerkship. Through using this approach in preclinical period, they can prepare themselves for the clinical environment to optimize the benefits derived from clerkships.


Assuntos
Estágio Clínico , Educação de Graduação em Medicina , Avaliação Educacional , Cirurgia Geral , Estudantes de Medicina , Estágio Clínico/métodos , Humanos , Masculino , Feminino , Educação de Graduação em Medicina/métodos , Cirurgia Geral/educação , Avaliação Educacional/métodos , Competência Clínica , Adulto Jovem
19.
West J Emerg Med ; 25(2): 209-212, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38596920

RESUMO

Introduction: Learners frequently benefit from modalities such as small-group, case-based teaching and interactive didactic experiences rather than passive learning methods. These contemporary techniques are features of Foundations of Emergency Medicine (FoEM) curricula, and particularly the Foundations I (F1) course, which targets first-year resident (PGY-1) learners. The American Board of Emergency Medicine administers the in-training exam (ITE) that provides an annual assessment of EM-specific medical knowledge. We sought to assess the effect of F1 implementation on ITE scores. Methods: We retrospectively analyzed data from interns at four EM residency programs accredited by the Accreditation Council for Graduate Medical Education. We collected data in 2021. Participating sites were geographically diverse and included three- and four-year training formats. We collected data from interns two years before (control group) and two years after (intervention group) implementation of F1 at each site. Year of F1 implementation ranged from 2015-2018 at participating sites. We abstracted data using a standard form including program, ITE raw score, year of ITE administration, US Medical Licensing Exam Step 1 score, Step 2 Clinical Knowledge (CK) score, and gender. We performed univariable and multivariable linear regression to explore differences between intervention and control groups. Results: We collected data for 180 PGY-1s. Step 1 and Step 2 CK scores were significant predictors of ITE in univariable analyses (both with P < 0.001). After accounting for Step 1 and Step 2 CK scores, we did not find F1 implementation to be a significant predictor of ITE score, P = 0.83. Conclusion: Implementation of F1 curricula did not show significant changes in performance on the ITE after controlling for important variables.


Assuntos
Medicina de Emergência , Internato e Residência , Humanos , Estados Unidos , Avaliação Educacional/métodos , Estudos Retrospectivos , Competência Clínica , Currículo , Medicina de Emergência/educação , Licenciamento em Medicina
20.
PLoS One ; 19(4): e0301728, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38603715

RESUMO

AR technology allows users to interact with virtual objects in real-world settings. Immersive AR experiences can enhance creativity and possibilities. Learners can explore real-life situations in a safe, controlled environment, understand abstract concepts and solve problems. This study investigates whether AR-codes affect boxing beginners' performance in some fundamental defensive techniques. An experimental and control group were created to implement a quasi-experimental design. By using the ASSURE instructional design model, AR technology was incorporated into the educational program and delivered in flipped classroom method to the experimental group. Comparatively, the control group is taught a program using a teacher's command style. A post-measurement of defensive boxing skills was conducted for both groups. Participants were 60 boxing beginners aged 12 to 14 who had enrolled in Port Fouad Sports Club's 2023/2024 training season in Port Said, Egypt. Randomly, participants were divided into control and experimental groups. They were homogenized and equivalent in terms of age, height, weight, IQ, physical fitness, and skill level. According to the study results, the experimental group performed better in post-measurements than the control group. The AR Codes technology had a large effect size on the learning of boxing defensive skills under study. Consequently, it is necessary to use AR Codes technology as an educational resource to enhance the learning process, integrate it with active learning strategies, and use it to teach defensive boxing skills and apply them to offensive and counterattack skills, thereby improving the learning process.


Assuntos
Realidade Aumentada , Boxe , Humanos , Aprendizagem Baseada em Problemas/métodos , Avaliação Educacional/métodos , Escolaridade
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA